A Multiclass Extension To The Brownboost Algorithm
نویسندگان
چکیده
Brownboost is an adaptive, continuous time boosting algorithm based on the Boost-by-Majority (BBM) algorithm. Though it has been little studied at the time of writing, it is believed that it should prove especially robust with respect to noisy data sets. This would make it a very useful boosting algorithm for real-world applications. More familiar algorithms such as Adaboost, or its successor Logitboost, are known to be especially susceptible to over tting the training data examples. This can lead to a poor generalization error in the presence of class noise, since weak hypotheses induced at later iterations to t the noisy examples will tend to be given undue in uence in the nal combined hypothesis. Brownboost allows us to specify an expected base-line error rate in advance, corresponding to our prior beliefs about the proportion of noise in the training data, and thus avoid over tting. The original derivation of Brownboost is restricted to binary classi cation problems. In this paper we propose a natural multi-class extension to the basic algorithm, incorporating error-correcting output codes and a multi-class gain measure. We test two-class and multi-class versions of the algorithm on a number of real and simulated data sets with articial class noise, and show that Brownboost consistently outperforms Adaboost in these situations.
منابع مشابه
Non-convex boosting with minimum margin guarantees
Many classification algorithms achieve poor generalization accuracy on “noisy” data sets. We introduce a new non-convex boosting algorithm BrownBoost-δ, a noiseresistant booster, that is able to significantly increase accuracy on a set of noisy classification problems. Our algorithm consistently outperforms the original BrownBoost algorithm, AdaBoost, and LogitBoost on simulated and real data. ...
متن کاملNon-Convex Boosting Overcomes Random Label Noise
The sensitivity of Adaboost to random label noise is a well-studied problem. LogitBoost, BrownBoost and RobustBoost are boosting algorithms claimed to be less sensitive to noise than AdaBoost. We present the results of experiments evaluating these algorithms on both synthetic and real datasets. We compare the performance on each of datasets when the labels are corrupted by different levels of i...
متن کاملOnline multiclass boosting
Recent work has extended the theoretical analysis of boosting algorithms to multiclass problems and to online settings. However, the multiclass extension is in the batch setting and the online extensions only consider binary classification. We fill this gap in the literature by defining, and justifying, a weak learning condition for online multiclass boosting. This condition leads to an optimal...
متن کاملRobust feature selection based on regularized brownboost loss
Feature selection is an important preprocessing step in machine learning and pattern recognition. It is also a data mining task in some real-world applications. Feature quality evaluation is a key issue when designing an algorithm for feature selection. The classification margin has been used widely to evaluate feature quality in recent years. In this study, we introduce a robust loss function,...
متن کاملAn Empirical Comparison of Three BoostingAlgorithms on Real Data Sets with Arti cialClass
Boosting algorithms are a means of building a strong ensemble classiier by aggregating a sequence of weak hypotheses. In this paper we consider three of the best-known boosting algorithms: Adaboost 8], Logitboost 10] and Brownboost 7]. These algorithms are adaptive, and work by maintaining a set of example and class weights which focus the attention of a base learner on the examples that are ha...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- IJPRAI
دوره 18 شماره
صفحات -
تاریخ انتشار 2004